46 research outputs found

    Best Linear Unbiased Estimation Fusion with Constraints

    Get PDF
    Estimation fusion, or data fusion for estimation, is the problem of how to best utilize useful information contained in multiple data sets for the purpose of estimating an unknown quantity — a parameter or a process. Estimation fusion with constraints gives rise to challenging theoretical problems given the observations from multiple geometrically dispersed sensors: Under dimensionality constraints, how to preprocess data at each local sensor to achieve the best estimation accuracy at the fusion center? Under communication bandwidth constraints, how to quantize local sensor data to minimize the estimation error at the fusion center? Under constraints on storage, how to optimally update state estimates at the fusion center with out-of-sequence measurements? Under constraints on storage, how to apply the out-of-sequence measurements (OOSM) update algorithm to multi-sensor multi-target tracking in clutter? The present work is devoted to the above topics by applying the best linear unbiased estimation (BLUE) fusion. We propose optimal data compression by reducing sensor data from a higher dimension to a lower dimension with minimal or no performance loss at the fusion center. For single-sensor and some particular multiple-sensor systems, we obtain the explicit optimal compression rule. For a multisensor system with a general dimensionality requirement, we propose the Gauss-Seidel iterative algorithm to search for the optimal compression rule. Another way to accomplish sensor data compression is to find an optimal sensor quantizer. Using BLUE fusion rules, we develop optimal sensor data quantization schemes according to the bit rate constraints in communication between each sensor and the fusion center. For a dynamic system, how to perform the state estimation and sensor quantization update simultaneously is also established, along with a closed form of a recursion for a linear system with additive white Gaussian noise. A globally optimal OOSM update algorithm and a constrained optimal update algorithm are derived to solve one-lag as well as multi-lag OOSM update problems. In order to extend the OOSM update algorithms to multisensor multitarget tracking in clutter, we also study the performance of OOSM update associated with the Probabilistic Data Association (PDA) algorithm

    Elastically-Constrained Meta-Learner for Federated Learning

    Full text link
    Federated learning is an approach to collaboratively training machine learning models for multiple parties that prohibit data sharing. One of the challenges in federated learning is non-IID data between clients, as a single model can not fit the data distribution for all clients. Meta-learning, such as Per-FedAvg, is introduced to cope with the challenge. Meta-learning learns shared initial parameters for all clients. Each client employs gradient descent to adapt the initialization to local data distributions quickly to realize model personalization. However, due to non-convex loss function and randomness of sampling update, meta-learning approaches have unstable goals in local adaptation for the same client. This fluctuation in different adaptation directions hinders the convergence in meta-learning. To overcome this challenge, we use the historical local adapted model to restrict the direction of the inner loop and propose an elastic-constrained method. As a result, the current round inner loop keeps historical goals and adapts to better solutions. Experiments show our method boosts meta-learning convergence and improves personalization without additional calculation and communication. Our method achieved SOTA on all metrics in three public datasets.Comment: FL-IJCAI'2

    Best Linear Unbiased Estimation Fusion with Constraints

    No full text
    Estimation fusion, or data fusion for estimation, is the problem of how to best utilize useful information contained in multiple data sets for the purpose of estimating an unknown quantity — a parameter or a process. Estimation fusion with constraints gives rise to challenging theoretical problems given the observations from multiple geometrically dispersed sensors: Under dimensionality constraints, how to preprocess data at each local sensor to achieve the best estimation accuracy at the fusion center? Under communication bandwidth constraints, how to quantize local sensor data to minimize the estimation error at the fusion center? Under constraints on storage, how to optimally update state estimates at the fusion center with out-of-sequence measurements? Under constraints on storage, how to apply the out-of-sequence measurements (OOSM) update algorithm to multi-sensor multi-target tracking in clutter? The present work is devoted to the above topics by applying the best linear unbiased estimation (BLUE) fusion. We propose optimal data compression by reducing sensor data from a higher dimension to a lower dimension with minimal or no performance loss at the fusion center. For single-sensor and some particular multiple-sensor systems, we obtain the explicit optimal compression rule. For a multisensor system with a general dimensionality requirement, we propose the Gauss-Seidel iterative algorithm to search for the optimal compression rule. Another way to accomplish sensor data compression is to find an optimal sensor quantizer. Using BLUE fusion rules, we develop optimal sensor data quantization schemes according to the bit rate constraints in communication between each sensor and the fusion center. For a dynamic system, how to perform the state estimation and sensor quantization update simultaneously is also established, along with a closed form of a recursion for a linear system with additive white Gaussian noise. A globally optimal OOSM update algorithm and a constrained optimal update algorithm are derived to solve one-lag as well as multi-lag OOSM update problems. In order to extend the OOSM update algorithms to multisensor multitarget tracking in clutter, we also study the performance of OOSM update associated with the Probabilistic Data Association (PDA) algorithm

    Efficient and Robust Feature Extraction by Maximum Margin Criterion

    No full text
    In pattern recognition, feature extraction techniques are widely employed to reduce the dimensionality of data and to enhance the discriminatory information. Principal Component Analysis (PCA) and Linear Discriminant Analysis (LDA) are two most popular linear dimen-sionality reduction methods. However, PCA is not very effective for the extraction of the most discriminant features and LDA is not stable due to the small sample size problem. In this pa-per, we propose some new (linear and nonlinear) feature extractors based on maximum margin criterion (MMC). Geometrically, feature extractors based on MMC maximize the (average) margin between classes after dimensionality reduction. It is shown that MMC can represent class separability better than PCA. As a connection to LDA, we may also derive LDA from MMC by incorporating some constraints. By using some other constraints, we establish a new linear feature extractor that does not suffer from the small sample size problem, which is known to cause serious stability problems for LDA. The kernelized (nonlinear) counterpart of this lin-ear feature extractor is also established in the paper. Our extensive experiments demonstrate that the new feature extractors are effective, stable, and efficient

    Minimum entropy clustering and applications to gene expression analysis. In:

    No full text
    Abstract Clustering is a common methodology for analyzing the gene expression data

    α-アルコキシ炭素ラジカルの分子間付加反応を用いた高酸化度炭素骨格の構築

    No full text
    学位の種別: 課程博士審査委員会委員 : (主査)東京大学教授 井上 将行, 東京大学教授 内山 真伸, 東京大学准教授 尾谷 優子, 東京大学講師 長友 優典, 東京大学講師 生長 幸之

    Relation of Optimal Local Compression and Local Likelihood Ratio

    No full text
    As Tenney and Sandell pointed out, the optimal local compression/decision rule has the form of a likelihood ratio when the local observations are not correlated [3]. However, this does not hold in general for correlated observations [4]. An interesting problem is to find the conditions under which the optimal local compression rule remains to have the form of a likelihood ratio even for correlated observations. In this paper, we prove that, with the model of Gaussian signal with independent Gaussian noise, the optimal local compression rule has the expected likelihood ratio form when the two conditional probability density functions are centrosymmetric. Computer simulation is provided in the paper for demonstration

    Optimal linear estimation fusion-part VI: Sensor data compression

    No full text
    Abstract – In many engineering applications, estimation accuracy can be improved by data from distributed sensors. Due to limited communication bandwidth and limited processing capability at the fusion center, it is crucial to compress these data for the final estimation at the fusion center. One way of accomplishing this is to reduce the dimension of the data with minimum or no loss of information. Based on the best linear unbiased estimation (BLUE) fusion results obtained in the previous parts of this series, in this paper we present optimal rules for compressing data at each local sensor to an allowable size (i.e., dimension) such that the fused estimate is optimal. We show that without any performance deterioration, all sensor data can be compressed to a dimension not larger than that of the estimatee (i.e., the quantity to be estimated). For some simple cases, these optimal compression rules are given analytically; for the general case, they can be found numerically by an algorithm proposed here. Supporting simulation results are provided
    corecore